60 research outputs found
A Model of Consistent Node Types in Signed Directed Social Networks
Signed directed social networks, in which the relationships between users can
be either positive (indicating relations such as trust) or negative (indicating
relations such as distrust), are increasingly common. Thus the interplay
between positive and negative relationships in such networks has become an
important research topic. Most recent investigations focus upon edge sign
inference using structural balance theory or social status theory. Neither of
these two theories, however, can explain an observed edge sign well when the
two nodes connected by this edge do not share a common neighbor (e.g., common
friend). In this paper we develop a novel approach to handle this situation by
applying a new model for node types. Initially, we analyze the local node
structure in a fully observed signed directed network, inferring underlying
node types. The sign of an edge between two nodes must be consistent with their
types; this explains edge signs well even when there are no common neighbors.
We show, moreover, that our approach can be extended to incorporate directed
triads, when they exist, just as in models based upon structural balance or
social status theory. We compute Bayesian node types within empirical studies
based upon partially observed Wikipedia, Slashdot, and Epinions networks in
which the largest network (Epinions) has 119K nodes and 841K edges. Our
approach yields better performance than state-of-the-art approaches for these
three signed directed networks.Comment: To appear in the IEEE/ACM International Conference on Advances in
Social Network Analysis and Mining (ASONAM), 201
Exemplar-Centered Supervised Shallow Parametric Data Embedding
Metric learning methods for dimensionality reduction in combination with
k-Nearest Neighbors (kNN) have been extensively deployed in many
classification, data embedding, and information retrieval applications.
However, most of these approaches involve pairwise training data comparisons,
and thus have quadratic computational complexity with respect to the size of
training set, preventing them from scaling to fairly big datasets. Moreover,
during testing, comparing test data against all the training data points is
also expensive in terms of both computational cost and resources required.
Furthermore, previous metrics are either too constrained or too expressive to
be well learned. To effectively solve these issues, we present an
exemplar-centered supervised shallow parametric data embedding model, using a
Maximally Collapsing Metric Learning (MCML) objective. Our strategy learns a
shallow high-order parametric embedding function and compares training/test
data only with learned or precomputed exemplars, resulting in a cost function
with linear computational complexity for both training and testing. We also
empirically demonstrate, using several benchmark datasets, that for
classification in two-dimensional embedding space, our approach not only gains
speedup of kNN by hundreds of times, but also outperforms state-of-the-art
supervised embedding approaches.Comment: accepted to IJCAI201
Localized Properties in Flakeboard: A Simulation Using Stacked Flakes
Heat transfer, vertical density distribution, bond strength, and dimensional stability were determined for columns of trembling aspen wood flakes pressed to simulate the density variation found within a flakeboard mat. Variables studied included: 1) the number of wood flakes in each column, 2) face flake moisture content, and 3) press closing time. The face, intermediate, and core layers of the resulting flake assemblies were evaluated in terms of the heat transfer occurring during pressing, their vertical density distribution, shear bond strength, and dimensional stability. More rapid heat transfer to the core of the flake assemblies was generally associated with shorter press closing times, higher moisture content face flakes, and lower initial numbers of flakes. Face densities were greatest for the shorter press closing time and low moisture content face flakes. Relative density differences between face and core layers were greatest for low numbers of flakes. Greatest strengths were found at the face layer and followed the vertical density distributions. Press closing time had no effect on strength. Face flake moisture content affected only the strength of the face and intermediate layers of the flake assembly composed of the greatest number of flakes. Thickness swelling trends closely followed the vertical density distributions
A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction
The Nonlinear autoregressive exogenous (NARX) model, which predicts the
current value of a time series based upon its previous values as well as the
current and past values of multiple driving (exogenous) series, has been
studied for decades. Despite the fact that various NARX models have been
developed, few of them can capture the long-term temporal dependencies
appropriately and select the relevant driving series to make predictions. In
this paper, we propose a dual-stage attention-based recurrent neural network
(DA-RNN) to address these two issues. In the first stage, we introduce an input
attention mechanism to adaptively extract relevant driving series (a.k.a.,
input features) at each time step by referring to the previous encoder hidden
state. In the second stage, we use a temporal attention mechanism to select
relevant encoder hidden states across all time steps. With this dual-stage
attention scheme, our model can not only make predictions effectively, but can
also be easily interpreted. Thorough empirical studies based upon the SML 2010
dataset and the NASDAQ 100 Stock dataset demonstrate that the DA-RNN can
outperform state-of-the-art methods for time series prediction.Comment: International Joint Conference on Artificial Intelligence (IJCAI),
201
Asymmetrical Hierarchical Networks with Attentive Interactions for Interpretable Review-Based Recommendation
Recently, recommender systems have been able to emit substantially improved
recommendations by leveraging user-provided reviews. Existing methods typically
merge all reviews of a given user or item into a long document, and then
process user and item documents in the same manner. In practice, however, these
two sets of reviews are notably different: users' reviews reflect a variety of
items that they have bought and are hence very heterogeneous in their topics,
while an item's reviews pertain only to that single item and are thus topically
homogeneous. In this work, we develop a novel neural network model that
properly accounts for this important difference by means of asymmetric
attentive modules. The user module learns to attend to only those signals that
are relevant with respect to the target item, whereas the item module learns to
extract the most salient contents with regard to properties of the item. Our
multi-hierarchical paradigm accounts for the fact that neither are all reviews
equally useful, nor are all sentences within each review equally pertinent.
Extensive experimental results on a variety of real datasets demonstrate the
effectiveness of our method
phonon anomaly driven by Fermi surface instability at intermediate temperature in YBaCuO
We performed temperature- and doping-dependent high-resolution Raman
spectroscopy experiments on YBaCuO to study
phonons. The temperature dependence of the real part of the phonon self-energy
shows a distinct kink at above due to softening,
in addition to the one due to the onset of the superconductivity. is clearly different from the pseudogap temperature with a maximum in the
underdoped region. The region between and
resembles that of superconducting fluctuation or charge density wave order.
While the true origin of the phonon softening is not known, we
can attribute it to a gap on the Fermi surface due to an electronic order. Our
results may reveal the role of the phonon not only in the
superconducting state but also in the intertwined orders in multilayer copper
oxide high- superconductors.Comment: 5 pages, 4 figure
Coffee: Boost Your Code LLMs by Fixing Bugs with Feedback
Code editing is an essential step towards reliable program synthesis to
automatically correct critical errors generated from code LLMs. Recent studies
have demonstrated that closed-source LLMs (i.e., ChatGPT and GPT-4) are capable
of generating corrective feedback to edit erroneous inputs. However, it remains
challenging for open-source code LLMs to generate feedback for code editing,
since these models tend to adhere to the superficial formats of feedback and
provide feedback with misleading information. Hence, the focus of our work is
to leverage open-source code LLMs to generate helpful feedback with correct
guidance for code editing. To this end, we present Coffee, a collected dataset
specifically designed for code fixing with feedback. Using this dataset, we
construct CoffeePots, a framework for COde Fixing with FEEdback via
Preference-Optimized Tuning and Selection. The proposed framework aims to
automatically generate helpful feedback for code editing while minimizing the
potential risk of superficial feedback. The combination of Coffee and
CoffeePots marks a significant advancement, achieving state-of-the-art
performance on HumanEvalFix benchmark. Codes and model checkpoints are publicly
available at https://github.com/Lune-Blue/COFFEE.Comment: Work in progres
Dynamic Gaussian Mixture based Deep Generative Model For Robust Forecasting on Sparse Multivariate Time Series
Forecasting on sparse multivariate time series (MTS) aims to model the
predictors of future values of time series given their incomplete past, which
is important for many emerging applications. However, most existing methods
process MTS's individually, and do not leverage the dynamic distributions
underlying the MTS's, leading to sub-optimal results when the sparsity is high.
To address this challenge, we propose a novel generative model, which tracks
the transition of latent clusters, instead of isolated feature representations,
to achieve robust modeling. It is characterized by a newly designed dynamic
Gaussian mixture distribution, which captures the dynamics of clustering
structures, and is used for emitting timeseries. The generative model is
parameterized by neural networks. A structured inference network is also
designed for enabling inductive analysis. A gating mechanism is further
introduced to dynamically tune the Gaussian mixture distributions. Extensive
experimental results on a variety of real-life datasets demonstrate the
effectiveness of our method.Comment: This paper is accepted by AAAI 202
Deep learning-based statistical noise reduction for multidimensional spectral data
In spectroscopic experiments, data acquisition in multi-dimensional phase
space may require long acquisition time, owing to the large phase space volume
to be covered. In such case, the limited time available for data acquisition
can be a serious constraint for experiments in which multidimensional spectral
data are acquired. Here, taking angle-resolved photoemission spectroscopy
(ARPES) as an example, we demonstrate a denoising method that utilizes deep
learning as an intelligent way to overcome the constraint. With readily
available ARPES data and random generation of training data set, we
successfully trained the denoising neural network without overfitting. The
denoising neural network can remove the noise in the data while preserving its
intrinsic information. We show that the denoising neural network allows us to
perform similar level of second-derivative and line shape analysis on data
taken with two orders of magnitude less acquisition time. The importance of our
method lies in its applicability to any multidimensional spectral data that are
susceptible to statistical noise.Comment: 8 pages, 8 figure
- …